75 research outputs found

    Asymmetric Totally-corrective Boosting for Real-time Object Detection

    Full text link
    Real-time object detection is one of the core problems in computer vision. The cascade boosting framework proposed by Viola and Jones has become the standard for this problem. In this framework, the learning goal for each node is asymmetric, which is required to achieve a high detection rate and a moderate false positive rate. We develop new boosting algorithms to address this asymmetric learning problem. We show that our methods explicitly optimize asymmetric loss objectives in a totally corrective fashion. The methods are totally corrective in the sense that the coefficients of all selected weak classifiers are updated at each iteration. In contract, conventional boosting like AdaBoost is stage-wise in that only the current weak classifier's coefficient is updated. At the heart of the totally corrective boosting is the column generation technique. Experiments on face detection show that our methods outperform the state-of-the-art asymmetric boosting methods.Comment: 14 pages, published in Asian Conf. Computer Vision 201

    Voltage island based heterogeneous NoC design through constraint programming

    Get PDF
    This paper discusses heterogeneous Network-on-Chip (NoC) design from a Constraint Programming (CP) perspective and extends the formulation to solving Voltage-Frequency Island (VFI) problem. In general, VFI is a superior design alternative in terms of thermal constraints, power consumption as well as performance considerations. Given a Communication Task Graph (CTG) and subsequent task assignments for cores, cores are allocated to the best possible places on the chip in the first stage to minimize the overall communication cost among cores. We then solve the application scheduling problem to determine the optimum core types from a list of technological alternatives and to minimize the makespan. Moreover, an elegant CP model is proposed to solve VFI problem by mapping and grouping cores at the same time with scheduling the computation tasks as a limited capacity resource allocation model. The paper reports results based on real benchmark datasets from the literature. © 2014 Elsevier Ltd. All rights reserved

    Confidence in prediction: an approach for dynamic weighted ensemble.

    Get PDF
    Combining classifiers in an ensemble is beneficial in achieving better prediction than using a single classifier. Furthermore, each classifier can be associated with a weight in the aggregation to boost the performance of the ensemble system. In this work, we propose a novel dynamic weighted ensemble method. Based on the observation that each classifier provides a different level of confidence in its prediction, we propose to encode the level of confidence of a classifier by associating with each classifier a credibility threshold, computed from the entire training set by minimizing the entropy loss function with the mini-batch gradient descent method. On each test sample, we measure the confidence of each classifier’s output and then compare it to the credibility threshold to determine whether a classifier should be attended in the aggregation. If the condition is satisfied, the confidence level and credibility threshold are used to compute the weight of contribution of the classifier in the aggregation. By this way, we are not only considering the presence but also the contribution of each classifier based on the confidence in its prediction on each test sample. The experiments conducted on a number of datasets show that the proposed method is better than some benchmark algorithms including a non-weighted ensemble method, two dynamic ensemble selection methods, and two Boosting methods

    Root canal morphology of primary maxillary second molars:a micro-computed tomography analysis

    Get PDF
    Aim Successful endodontic treatment of primary teeth requires comprehensive knowledge and understanding of root canal morphology. The purpose of this study was to investigate the root canal configurations of primary maxillary second molars using micro-computed tomography. Methods Extracted human primary maxillary second molars (n = 57) were scanned using micro-computed tomography and reconstructed to produce three-dimensional models. Each root canal system was analysed qualitatively according to Vertucci's classification. Results 22.8% (n = 13) of the sample presented with the fusion of the disto-buccal and palatal roots; of these, Type V was the most prevalent classification. For teeth with three separate roots (n = 44), the most common root canal type was Type 1 for the palatal canal (100%) and disto-buccal canal (77.3%) and Type V for the mesio-buccal canal (36.4%). Overall, 7% (n = 4) of mesio-buccal canals were 'unclassifiable'. Conclusion The root canal systems of primary maxillary second molars were not only complex but had a range of configurations that may contribute to unfavourable clinical outcomes after endodontic treatment

    Automated Ham Quality Classification Using Ensemble Unsupervised Mapping Models

    Get PDF
    This multidisciplinary study focuses on the application and comparison of several topology preserving mapping models upgraded with some classifier ensemble and boosting techniques in order to improve those visualization capabilities. The aim is to test their suitability for classification purposes in the field of food industry and more in particular in the case of dry cured ham. The data is obtained from an electronic device able to emulate a sensory olfative taste of ham samples. Then the data is classified using the previously mentioned techniques in order to detect which batches have an anomalous smelt (acidity, rancidity and different type of taints) in an automated way

    Binarized Support Vector Machines

    Get PDF
    The widely used support vector machine (SVM) method has shown to yield very good results in supervised classification problems. Other methods such as classification trees have become more popular among practitioners than SVM thanks to their interpretability, which is an important issue in data mining. In this work, we propose an SVM-based method that automatically detects the most important predictor variables and the role they play in the classifier. In particular, the proposed method is able to detect those values and intervals that are critical for the classification. The method involves the optimization of a linear programming problem in the spirit of the Lasso method with a large number of decision variables. The numerical experience reported shows that a rather direct use of the standard column generation strategy leads to a classification method that, in terms of classification ability, is competitive against the standard linear SVM and classification trees. Moreover, the proposed method is robust; i.e., it is stable in the presence of outliers and invariant to change of scale or measurement units of the predictor variables. When the complexity of the classifier is an important issue, a wrapper feature selection method is applied, yielding simpler but still competitive classifiers

    Statistical strategies for avoiding false discoveries in metabolomics and related experiments

    Full text link
    corecore